Penalized Regression Methods for Linear Models in SAS/STAT
نویسنده
چکیده
Regression problems with many potential candidate predictor variables occur in a wide variety of scientific fields and business applications. These problems require you to perform statistical model selection to find an optimal model, one that is as simple as possible while still providing good predictive performance. Traditional stepwise selection methods, such as forward and backward selection, suffer from high variability and low prediction accuracy, especially when there are many predictor variables or correlated predictor variables (or both). In the last decade, the higher prediction accuracy and computational efficiency of penalized regression methods have made them an attractive alternative to traditional selection methods. This paper first provides a brief review of the LASSO, adaptive LASSO, and elastic net penalized model selection methods. Then it explains how to perform model selection by applying these techniques with the GLMSELECT procedure, which includes extensive customization options and powerful graphs for steering statistical model selection.
منابع مشابه
Penalized Bregman Divergence Estimation via Coordinate Descent
Variable selection via penalized estimation is appealing for dimension reduction. For penalized linear regression, Efron, et al. (2004) introduced the LARS algorithm. Recently, the coordinate descent (CD) algorithm was developed by Friedman, et al. (2007) for penalized linear regression and penalized logistic regression and was shown to gain computational superiority. This paper explores...
متن کاملPenalized Estimators in Cox Regression Model
The proportional hazard Cox regression models play a key role in analyzing censored survival data. We use penalized methods in high dimensional scenarios to achieve more efficient models. This article reviews the penalized Cox regression for some frequently used penalty functions. Analysis of medical data namely ”mgus2” confirms the penalized Cox regression performs better than the cox regressi...
متن کاملRobust Estimation in Linear Regression with Molticollinearity and Sparse Models
One of the factors affecting the statistical analysis of the data is the presence of outliers. The methods which are not affected by the outliers are called robust methods. Robust regression methods are robust estimation methods of regression model parameters in the presence of outliers. Besides outliers, the linear dependency of regressor variables, which is called multicollinearity...
متن کاملFixing the nonconvergence bug in logistic regression with SPLUS and SAS
When analyzing clinical data with binary outcomes, the parameter estimates and consequently the odds ratio estimates of a logistic model sometimes do not converge to finite values. This phenomenon is due to special conditions in a data set and known as 'separation'. Statistical software packages for logistic regression using the maximum likelihood method cannot appropriately deal with this prob...
متن کاملAnalyzing a Regression Model with a General Positive Definite Covariance Matrix with The SAS System
This article discusses and proposes a procedure for the analysis of the univariate linear regression model with known general positive definite covariance matrix with SAS/STAT software of the SAS System. Estimation of parameters, hypothesis testing, estimation under constraints and collinearity and influence diagnostics are reviewed. An example is given to illustrate the procedure.
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2015